Dif-MAML: Decentralized Multi-Agent Meta-Learning
نویسندگان
چکیده
The objective of meta-learning is to exploit the knowledge obtained from observed tasks improve adaptation unseen tasks. As such, meta-learners are able generalize better when they trained with a larger number and amount data per task. Given resources that needed, it generally difficult expect tasks, their respective data, necessary computational capacity be available at single central location. It more natural encounter situations where these spread across several agents connected by some graph topology. formalism actually well-suited this decentralized setting, learner would benefit information power agents. Motivated observation, in work, we propose cooperative fully-decentralized multi-agent algorithm, referred as Diffusion-based MAML or Dif-MAML. Decentralized optimization algorithms superior centralized implementations terms scalability, avoidance communication bottlenecks, privacy guarantees. work provides detailed theoretical analysis show proposed strategy allows collection attain agreement linear rate converge stationary point aggregate even non-convex environments. Simulation results illustrate findings performance relative traditional non-cooperative setting.
منابع مشابه
Decentralized Anti-coordination Through Multi-agent Learning
To achieve an optimal outcome in many situations, agents need to choose distinct actions from one another. This is the case notably in many resource allocation problems, where a single resource can only be used by one agent at a time. How shall a designer of a multi-agent system program its identical agents to behave each in a different way? From a game theoretic perspective, such situations le...
متن کاملCoordinated Multi-Agent Learning for Decentralized POMDPs
In many multi-agent applications such as distributed sensor nets, a network of agents act collaboratively under uncertainty and local interactions. Networked Distributed POMDP (ND-POMDP) provides a framework to model such cooperative multi-agent decision making. Existing work on ND-POMDPs has focused on offline techniques that require accurate models, which are usually costly to obtain in pract...
متن کاملDynamic Safe Interruptibility for Decentralized Multi-Agent Reinforcement Learning
In reinforcement learning, agents learn by performing actions and observing their 1 outcomes. Sometimes, it is desirable for a human operator to interrupt an agent 2 in order to prevent dangerous situations from happening. Yet, as part of their 3 learning process, agents may link these interruptions, that impact their reward, to 4 specific states and deliberately avoid them. The situation is pa...
متن کاملFully Decentralized Multi-Agent Reinforcement Learning with Networked Agents
We consider the problem of fully decentralized multi-agent reinforcement learning (MARL), where the agents are located at the nodes of a time-varying communication network. Specifically, we assume that the reward functions of the agents might correspond to different tasks, and are only known to the corresponding agent. Moreover, each agent makes individual decisions based on both the informatio...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE open journal of signal processing
سال: 2022
ISSN: ['2644-1322']
DOI: https://doi.org/10.1109/ojsp.2021.3140000